6 research outputs found

    IKH-EFT: An improved method of workflow scheduling using the krill herd algorithm in the fog-cloud environment

    Get PDF
    peer reviewedGiven the increase diversity of smart devices and objectives of the application management such as energy consumption, makespan users expect their requests to be responded to in an appropriate computation environment as properly as possible. In this paper, a method of workflow scheduling based on the fog-cloud architecture has been designed given the high processing capability of the cloud and the close communication between the user and the fog computing node, which reduces delay in response. We also seek to minimize consumption and reduce energy use and monetary cost in order to maximize customer satisfaction with proper scheduling. Given the large number of variables that are used in workflow scheduling and the optimization of contradictory objectives, the problem is NP-hard, and the multi-objective metaheuristic krill herd algorithm is used to solve it. The initial population is generated in a smart fashion to allow fast convergence of the algorithm. For allocation of tasks to the available fog-cloud resources, the EFT (earliest finish time) technique is used, and resource voltage and frequency are assumed to be dynamic to reduce energy use. A comprehensive simulation has been made for assessment of the proposed method in different scenarios with various values of CCR. The simulation results indicate that makespan exhibits improvements by 9.9, 8.7% and 6.7% on average compared with respect to the methods of IHEFT, HEFT and IWO-CA, respectively. Moreover, the monetary cost of the method and energy use have simultaneously decreased in the fog-cloud environment

    TrustDL: Use of trust-based dictionary learning to facilitate recommendation in social networks

    Get PDF
    peer reviewedCollaborative filtering (CF) is a widely applied method to perform recommendation tasks in a wide range of domains and applications. Dictionary learning (DL) models, which are highly important in CF-based recommender systems (RSs), are well represented by rating matrices. However, these methods alone do not resolve the cold start and data sparsity issues in RSs. We observed a significant improvement in rating results by adding trust information on the social network. For that purpose, we proposed a new dictionary learning technique based on trust information, called TrustDL, where the social network data were employed in the process of recommendation based on structural details on the trusted network. TrustDL sought to integrate the sources of information, including trust statements and ratings, into the recommendation model to mitigate both problems of cold start and data sparsity. It conducted dictionary learning and trust embedding simultaneously to predict unknown rating values. In this paper, the dictionary learning technique was integrated into rating learning, along with the trust consistency regularization term designed to offer a more accurate understanding of the feature representation. Moreover, partially identical trust embedding was developed, where users with similar rating sets could cluster together, and those with similar rating sets could be represented collaboratively. The proposed strategy appears significantly beneficial based on experiments conducted on four frequently used datasets: Epinions, Ciao, FilmTrust, and Flixster

    An energy-efficient and deadline-aware workflow scheduling algorithm in the fog and cloud environment

    Get PDF
    peer reviewedThe Internet of Things (IoT) is constantly evolving. The variety of IoT applications has caused new demands to emerge on users’ part and competition between computing service providers. On the one hand, an IoT application may exhibit several important criteria, such as deadline and runtime simultaneously, and it is confronted with resource limitations and high energy consumption on the other hand. This has turned to adopting a computing environment and scheduling as a fundamental challenge. To resolve the issue, IoT applications are considered in this paper as a workflow composed of a series of interdependent tasks. The tasks in the same workflow (at the same level) are subject to priorities and deadlines for execution, making the problem far more complex and closer to the real world. In this paper, a hybrid Particle Swarm Optimization and Simulated Annealing algorithm (PSO–SA) is used for prioritizing tasks and improving fitness function. Our proposed method managed the task allocation and optimized energy consumption and makespan at the fog-cloud environment nodes. The simulation results indicated that the PSO–SA enhanced energy and makespan by 5% and 9% respectively on average compared with the baseline algorithm (IKH-EFT)

    A distributed computing model for big data anonymization in the networks.

    No full text
    Recently big data and its applications had sharp growth in various fields such as IoT, bioinformatics, eCommerce, and social media. The huge volume of data incurred enormous challenges to the architecture, infrastructure, and computing capacity of IT systems. Therefore, the compelling need of the scientific and industrial community is large-scale and robust computing systems. Since one of the characteristics of big data is value, data should be published for analysts to extract useful patterns from them. However, data publishing may lead to the disclosure of individuals' private information. Among the modern parallel computing platforms, Apache Spark is a fast and in-memory computing framework for large-scale data processing that provides high scalability by introducing the resilient distributed dataset (RDDs). In terms of performance, Due to in-memory computations, it is 100 times faster than Hadoop. Therefore, Apache Spark is one of the essential frameworks to implement distributed methods for privacy-preserving in big data publishing (PPBDP). This paper uses the RDD programming of Apache Spark to propose an efficient parallel implementation of a new computing model for big data anonymization. This computing model has three-phase of in-memory computations to address the runtime, scalability, and performance of large-scale data anonymization. The model supports partition-based data clustering algorithms to preserve the λ-diversity privacy model by using transformation and actions on RDDs. Therefore, the authors have investigated Spark-based implementation for preserving the λ-diversity privacy model by two designed City block and Pearson distance functions. The results of the paper provide a comprehensive guideline allowing the researchers to apply Apache Spark in their own researches
    corecore